Slope-adaptive Variable Selection via Convex Optimization.
نویسندگان
چکیده
We introduce a new estimator for the vector of coefficients β in the linear model y = Xβ + z, where X has dimensions n × p with p possibly larger than n. SLOPE, short for Sorted L-One Penalized Estimation, is the solution to [Formula: see text]where λ1 ≥ λ2 ≥ … ≥ λ p ≥ 0 and [Formula: see text] are the decreasing absolute values of the entries of b. This is a convex program and we demonstrate a solution algorithm whose computational complexity is roughly comparable to that of classical ℓ1 procedures such as the Lasso. Here, the regularizer is a sorted ℓ1 norm, which penalizes the regression coefficients according to their rank: the higher the rank-that is, stronger the signal-the larger the penalty. This is similar to the Benjamini and Hochberg [J. Roy. Statist. Soc. Ser. B57 (1995) 289-300] procedure (BH) which compares more significant p-values with more stringent thresholds. One notable choice of the sequence {λ i } is given by the BH critical values [Formula: see text], where q ∈ (0, 1) and z(α) is the quantile of a standard normal distribution. SLOPE aims to provide finite sample guarantees on the selected model; of special interest is the false discovery rate (FDR), defined as the expected proportion of irrelevant regressors among all selected predictors. Under orthogonal designs, SLOPE with λBH provably controls FDR at level q. Moreover, it also appears to have appreciable inferential properties under more general designs X while having substantial power, as demonstrated in a series of experiments running on both simulated and real data.
منابع مشابه
Group SLOPE – Adaptive Selection of Groups of Predictors
Sorted L-One Penalized Estimation (SLOPE, [10]) is a relatively new convex optimization procedure which allows for adaptive selection of regressors under sparse high dimensional designs. Here we extend the idea of SLOPE to deal with the situation when one aims at selecting whole groups of explanatory variables instead of single regressors. This approach is particularly useful when variables in ...
متن کاملSlope—adaptive Variable Selection via Convex Optimization by Małgorzata Bogdan∗,1, Ewout
min b∈Rp 1 2 ‖y −Xb‖2 2 + λ1|b|(1) + λ2|b|(2) + · · · + λp|b|(p), where λ1 ≥ λ2 ≥ · · · ≥ λp ≥ 0 and |b|(1) ≥ |b|(2) ≥ · · · ≥ |b|(p) are the decreasing absolute values of the entries of b. This is a convex program and we demonstrate a solution algorithm whose computational complexity is roughly comparable to that of classical 1 procedures such as the Lasso. Here, the regularizer is a sorted 1 ...
متن کاملRejoinder: Latent Variable Graphical Model Selection via Convex Optimization by Venkat Chandrasekaran,
متن کامل
Slope Stability Analysis Using a Self-Adaptive Genetic Algorithm
This paper introduces a methodology for soil slope stability analysis based on optimization, limit equilibrium principles and method of slices. In this study, the slope stability analysis problem is transformed into a constrained nonlinear optimization problem. To solve that, a Self-Adaptive Genetic Algorithm (GA) is utilized. In this study, the slope stability safety factors are the objective ...
متن کاملFaster Coordinate Descent via Adaptive Importance Sampling
Coordinate descent methods employ random partial updates of decision variables in order to solve huge-scale convex optimization problems. In this work, we introduce new adaptive rules for the random selection of their updates. By adaptive, we mean that our selection rules are based on the dual residual or the primal-dual gap estimates and can change at each iteration. We theoretically character...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- The annals of applied statistics
دوره 9 3 شماره
صفحات -
تاریخ انتشار 2015